Small-Vote Sample Selection for Label-Noise Learning
نویسندگان
چکیده
The small-loss criterion is widely used in recent label-noise learning methods. However, such a only considers the loss of each training sample mini-batch but ignores distribution whole set. Moreover, selection clean samples depends on heuristic data rate. As result, some noisy-labeled are easily identified as ones, and vice versa. In this paper, we propose novel yet simple method, which mainly consists Hierarchical Voting Scheme (HVS) an Adaptive Clean rate Estimation Strategy (ACES), to accurately identify for robust learning. Specifically, HVS effectively combine global vote local vote, so that both epoch-level batch-level information exploited assign hierarchical sample. Based HVS, further develop ACES adaptively estimate by leveraging 1D Gaussian Mixture Model (GMM). Experimental results show our proposed method consistently outperforms several state-of-the-art methods synthetic real-world noisy benchmark datasets.
منابع مشابه
An Effective Approach for Robust Metric Learning in the Presence of Label Noise
Many algorithms in machine learning, pattern recognition, and data mining are based on a similarity/distance measure. For example, the kNN classifier and clustering algorithms such as k-means require a similarity/distance function. Also, in Content-Based Information Retrieval (CBIR) systems, we need to rank the retrieved objects based on the similarity to the query. As generic measures such as ...
متن کاملFeature Selection for Small Sample Sets with High Dimensional Data Using Heuristic Hybrid Approach
Feature selection can significantly be decisive when analyzing high dimensional data, especially with a small number of samples. Feature extraction methods do not have decent performance in these conditions. With small sample sets and high dimensional data, exploring a large search space and learning from insufficient samples becomes extremely hard. As a result, neural networks and clustering a...
متن کاملFeature Selection in Distance Learning from Small Sample
Learning from a small sample is an acute problem which arises in many applications where acquiring new samples is difficult, time consuming or expensive. The problem becomes even harder when dealing with rich high dimensional data. The learning process in such cases is often preceded by dimensionality reduction or feature selection. The need to avoid overfitting of an algorithm to the data is c...
متن کاملFeature selection for multi-label learning
Feature Selection plays an important role in machine learning and data mining, and it is often applied as a data pre-processing step. This task can speed up learning algorithms and sometimes improve their performance. In multi-label learning, label dependence is considered another aspect that can contribute to improve learning performance. A replicable and wide systematic review performed by us...
متن کاملExploiting Multi-Label Information for Noise Resilient Feature Selection
In conventional supervised learning paradigm, each data instance is associated with one single class label. Multi-label learning differs in the way that data instances may belong to multiple concepts simultaneously, which naturally appear in a variety of high impact domains, ranging from bioinformatics, information retrieval to multimedia analysis. It targets to leverage the multiple label info...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2021
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-030-86523-8_44